The Pennsylvania State University The Graduate School TWO TOPICS: A JACKKNIFE MAXIMUM LIKELIHOOD APPROACH TO STATISTICAL MODEL SELECTION AND A CONVEX HULL PEELING DEPTH APPROACH TO NONPARAMETRIC MASSIVE MULTIVARIATE DATA ANALYSIS WITH APPLICATIONS

نویسندگان

  • Jogesh Babu
  • Thomas P. Hettmansperger
  • Bing Li
  • James L. Rosenberger
چکیده

This dissertation presents two topics from opposite disciplines: one is from a parametric realm and the other is based on nonparametric methods. The first topic is a jackknife maximum likelihood approach to statistical model selection and the second one is a convex hull peeling depth approach to nonparametric massive multivariate data analysis. The second topic includes simulations and applications on massive astronomical data. First, we present a model selection criterion, minimizing the Kullback-Leibler distance by using the jackknife method. Various model selection methods have been developed to choose a model of minimum Kullback-Liebler distance to the true model, such as Akaike information criterion (AIC), Bayesian information criterion (BIC), Minimum description length (MDL), and Bootstrap information criterion. Likewise, the jackknife method chooses a model of minimum KullbackLeibler distance through bias reduction. This bias, which is inevitable in model selection problems, arise from estimating the distance between an unknown true model and an estimated model. We show that (a) the jackknife maximum likelihood estimator is consistent to the parameter of interest, (b) the jackknife estimate of the log likelihood is asymptotically unbiased, and (c) the stochastic order of the jackknife log likelihood estimate is O(log log n). Because of these properties, the jackknife information criterion is applicable to problems of choosing a model from non nested candidates especially when the true model is unknown. Compared to popular information criteria which are only applicable to nested models, the jackknife information criterion is more robust in terms of filtering various types of candidate models to choose the best approximating model. However, this robust method has a demerit that the jackknife criterion is unable to discriminate nested models. Next, we explore the convex hull peeling process to develop empirical tools for statistical inferences on multivariate massive data. Convex hull and its peeling

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Flood Flow Frequency Model Selection Using L-moment Method in Arid and Semi Arid Regions of Iran

Statistical frequency analysis is the most common procedure for the analysis of flood data at a gauged location thatin first step it is needed to select a model to represent the population. Among them, the central moment has been themost common and widely used, and with the using of computers, the application of the maximum likelihood hasincreased. This research was carried out in order to reco...

متن کامل

Fuzzy-rough Information Gain Ratio Approach to Filter-wrapper Feature Selection

Feature selection for various applications has been carried out for many years in many different research areas. However, there is a trade-off between finding feature subsets with minimum length and increasing the classification accuracy. In this paper, a filter-wrapper feature selection approach based on fuzzy-rough gain ratio is proposed to tackle this problem. As a search strategy, a modifie...

متن کامل

Sustainable Supplier Selection: A New Integrated Approach of Fuzzy Interpretive Structural Modeling and Dynamic Network Data Envelopment Analysis

Data envelopment analysis (DEA), as a well-established nonparametric method, is used to meet efficiency evaluation purposes in many businesses, organizations, and decision units. This paper aims to present a novel integrated approach to fuzzy interpretive structural modeling (FISM) and dynamic network data envelopment analysis (DNDEA) for the selection and ranking of sustainable suppliers. Firs...

متن کامل

Statistical Inference in Autoregressive Models with Non-negative Residuals

Normal residual is one of the usual assumptions of autoregressive models but in practice sometimes we are faced with non-negative residuals case. In this paper we consider some autoregressive models with non-negative residuals as competing models and we have derived the maximum likelihood estimators of parameters based on the modified approach and EM algorithm for the competing models. Also,...

متن کامل

Differenced-Based Double Shrinking in Partial Linear Models

Partial linear model is very flexible when the relation between the covariates and responses, either parametric and nonparametric. However, estimation of the regression coefficients is challenging since one must also estimate the nonparametric component simultaneously. As a remedy, the differencing approach, to eliminate the nonparametric component and estimate the regression coefficients, can ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1989